Opponent Modeling against Non-stationary Strategies: (Doctoral Consortium)

نویسندگان

  • Pablo Hernandez-Leal
  • Enrique Munoz de Cote
  • Luis Enrique Sucar
چکیده

Most state of the art learning algorithms do not fare well with agents (computer or humans) that change their behaviour in time. This is the case because they usually do not model the other agents’ behaviour and instead make some assumptions that for real scenarios are too restrictive. Furthermore, considering that many applications demand different types of agents to work together this should be an important problem to solve. We contribute to the state of the art with opponent modeling algorithms. In particular we proposed 3 approaches for learning against non-stationary opponents in repeated games. Experimentally we tested our approaches on three domains including a real world scenario which consists of bidding in energy markets.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Using a Priori Information for Fast Learning Against Non-stationary Opponents

For an agent to be successful in interacting against many different and unknown types of opponents it should excel at learning fast a model of the opponent and adapt online to non-stationary (changing) strategies. Recent works have tackled this problem by continuously learning models of the opponent while checking for switches in the opponent strategy. However, these approaches fail to use a pr...

متن کامل

Modeling non-stationary opponents

This paper studies repeated interactions between an agent and an unknown opponent that changes its strategy over time. We propose a framework for learning switching nonstationary strategies. The approach uses decision trees to learn the most up to date opponent’s strategy. Then, the agent’s strategy is computed by transforming the tree into a Markov Decision Process (MDP), whose solution dictat...

متن کامل

Combining Opponent Modeling and Model-Based Reinforcement Learning in a Two-Player Competitive Game

When an opponent with a stationary and stochastic policy is encountered in a twoplayer competitive game, model-free Reinforcement Learning (RL) techniques such as Q-learning and Sarsa(λ) can be used to learn near-optimal counter strategies given enough time. When an agent has learned such counter strategies against multiple diverse opponents, it is not trivial to decide which one to use when a ...

متن کامل

Learning Against Non-Stationary Opponents in Double Auctions

Energy markets are emerging around the world. In this context, the PowerTAC competition has gained attention for being a realistic and powerful simulation platform that can be used to perform robust research on retail energy markets. Agent in this complex environment typically use different strategies throughout their interaction, changing from one to another depending on diverse factors, for e...

متن کامل

An Experimental Approach to Online Opponent Modeling in Texas Hold'em Poker

The game of Poker is an excellent test bed for studying opponent modeling methodologies applied to non-deterministic games with incomplete information. The most known Poker variant, Texas Hold'em Poker, combines simple rules with a huge amount of possible playing strategies. This paper is focused on developing algorithms for performing simple online opponent modeling in Texas Hold'em. The oppon...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2015